# Wikipedia training
Yugogpt Florida Q8 0 GGUF
MIT
YugoGPT-Florida is a large language model based on Serbian, excelling in multiple evaluation benchmarks.
Large Language Model Other
Y
MarkoRadojcic
30
2
Tiny Lm
MIT
A small language model with only 16 million parameters, mainly used for debugging and testing, supporting English and Japanese.
Large Language Model
Transformers Supports Multiple Languages

T
sbintuitions
6,626
2
Simcse Model XLMR
Apache-2.0
A sentence-transformers model based on XLM-R, trained using the SimCSE method, which maps sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as clustering or semantic search.
Text Embedding
Transformers

S
kornwtp
20
0
Simcse Model Phayathaibert
Apache-2.0
This is a model based on sentence-transformers that can map sentences and paragraphs into a 768-dimensional dense vector space, suitable for tasks such as clustering or semantic search.
Text Embedding
Transformers

S
kornwtp
123
2
Simcse Model M Bert Thai Cased
Apache-2.0
A SimCSE model based on mBERT, specifically trained for Thai language, used to generate 768-dimensional vector representations of sentences and paragraphs
Text Embedding
Transformers

S
kornwtp
25
1
Minilm L6 H384 Italian Cross Encoder
MIT
Italian text ranking model based on MiniLMv2 architecture, with optimized embedding layers for Italian
Text Embedding Other
M
osiria
328
1
Abstract Sim Query
A model that maps abstract sentence descriptions to matching sentences, trained on Wikipedia using a dual-encoder architecture.
Text Embedding
Transformers English

A
biu-nlp
53
12
Abstract Sim Sentence
A model that maps abstract sentence descriptions to matching sentences, trained on Wikipedia using a dual encoder architecture.
Text Embedding
Transformers English

A
biu-nlp
51
16
Sbert Large Cased Pl
SHerbert large is an improved SentenceBERT model based on the Polish HerBERT, designed to generate semantically meaningful sentence embeddings and compare them using cosine similarity.
Text Embedding Other
S
Voicelab
327
7
Multilingual Bert Gn Base Cased
MIT
A language model fine-tuned for Guarani based on the multilingual BERT base model, supporting 104 languages including Guarani
Large Language Model
Transformers Other

M
mmaguero
28
0
Dansk Gpt Wiki
A Danish GPT2-style model trained using the Flax CLM pipeline, based on the Danish portion of the wiki40b dataset.
Large Language Model Other
D
flax-community
14.89k
4
Bert Base Mongolian Cased
This is a pre-trained Mongolian BERT model, trained on Mongolian Wikipedia and news datasets, supporting Mongolian text processing tasks.
Large Language Model Other
B
tugstugi
258
0
Simcse Model Distil M Bert
A sentence transformer model based on m-Distil-BERT, trained using the SimCSE method, capable of mapping text to 768-dimensional vectors, suitable for semantic search and clustering tasks
Text Embedding
Transformers

S
mrp
21
0
Bertinho Gl Small Cased
A pre-trained BERT model for Galician (6 layers, case-sensitive). Trained on Wikipedia.
Large Language Model Other
B
dvilares
56
2
Swe Gpt Wiki
This is a Swedish GPT2-style model trained using the Flax CLM process, with training data from the Swedish portion of the wiki40b dataset.
Large Language Model Other
S
flax-community
24
3
Dbert
A Korean pre-trained language model based on the BERT architecture, suitable for Korean text processing tasks.
Large Language Model
Transformers Korean

D
baikalai
17
0
Simcse Model M Bert Thai Cased
A Thai sentence embedding model based on mBERT, trained using the SimCSE method on Thai Wikipedia data, capable of mapping text to 768-dimensional vectors
Text Embedding
Transformers

S
mrp
1,637
7
Bert Base Thai Upos
Apache-2.0
BERT model pre-trained on Thai Wikipedia text for POS tagging and dependency parsing
Sequence Labeling
Transformers Other

B
KoichiYasuoka
53.03k
1
Bert Base It Cased
Apache-2.0
This is a customized streamlined version of bert-base-multilingual-cased, specifically optimized for Italian language processing while maintaining the original model's accuracy.
Large Language Model Other
B
Geotrend
15
0
Distilbert Base En Vi Cased
Apache-2.0
This is a compact version of distilbert-base-multilingual-cased, specifically designed for English and Vietnamese, maintaining the original model's accuracy.
Large Language Model
Transformers Other

D
Geotrend
30
1
Bert Base En Ja Cased
Apache-2.0
A compact version customized from bert-base-multilingual-cased, focusing on English and Japanese processing while maintaining the original model's representational capabilities.
Large Language Model Other
B
Geotrend
749
0
Distilbert Base En Fr Da Ja Vi Cased
Apache-2.0
This is a lightweight version of distilbert-base-multilingual-cased, supporting English, French, Danish, Japanese, and Vietnamese processing while maintaining the original model's accuracy.
Large Language Model
Transformers Other

D
Geotrend
25
0
Distilbert Base Ur Cased
Apache-2.0
This is a lightweight version of distilbert-base-multilingual-cased, specifically optimized for Urdu, maintaining the accuracy of the original model.
Large Language Model
Transformers Other

D
Geotrend
157
0
Distilbert Base En Zh Cased
Apache-2.0
This is a compact version of distilbert-base-multilingual-cased, specifically designed for bilingual tasks in English and Chinese, maintaining the accuracy of the original model.
Large Language Model
Transformers Other

D
Geotrend
29
1
Featured Recommended AI Models